# Large Language Model
Josiefied Qwen3 30B A3B Abliterated V2 4bit
This is a 4-bit quantized version converted from the Qwen3-30B model, suitable for text generation tasks on the MLX framework.
Large Language Model
J
mlx-community
194
1
Deepseek R1 0528 Qwen3 8B Bf16
MIT
This model is an MLX format version converted from deepseek-ai/deepseek-r1-0528-Qwen3-8B, suitable for local inference on Apple devices.
Large Language Model
D
mlx-community
2,298
1
PKU DS LAB.FairyR1 32B GGUF
FairyR1-32B is a large language model with 32B parameters, developed by PKU-DS-LAB, focusing on text generation tasks.
Large Language Model
P
DevQuasar
134
1
Qwen3 32B 4bit DWQ
Apache-2.0
Qwen3-32B-8bit is an 8-bit quantized version of the Qwen3-32B model, designed for text generation tasks and released by mlx-community.
Large Language Model
Q
mlx-community
211
1
Qwen3 235B A22B 4bit DWQ
Apache-2.0
Qwen3-235B-A22B-4bit-DWQ is a 4-bit quantized version converted from the Qwen3-235B-A22B-8bit model, suitable for text generation tasks.
Large Language Model
Q
mlx-community
70
1
Avern 1.5 Mintra
MIT
Qwen2.5-Coder-7B-Instruct is a 7B-parameter code generation model based on the Qwen2.5 architecture, specializing in instruction fine-tuning, suitable for code generation and programming assistance tasks.
Large Language Model
A
averntech
87
1
Qwen3 235B A22B Mixed 3 6bit
Apache-2.0
This is a mixed 3-6bit quantized version converted from the Qwen/Qwen3-235B-A22B model, optimized for efficient inference on the Apple MLX framework.
Large Language Model
Q
mlx-community
100
2
Qwen Qwen2.5 VL 72B Instruct GGUF
Other
A quantized version of the Qwen2.5-VL-72B-Instruct multimodal large language model, supporting image-text-to-text tasks, suitable for various quantization levels from high precision to low memory requirements.
Text-to-Image English
Q
bartowski
1,336
1
Qwen3 30B A3B MNN
Apache-2.0
An MNN model exported from Qwen3-30B-A3B, featuring 4-bit quantization for efficient inference.
Large Language Model English
Q
taobao-mnn
550
1
Qwen3 30B A3B 4bit DWQ
Apache-2.0
This is a 4-bit quantized version based on the Qwen3-30B-A3B model, created through custom DWQ quantization technology distilled from 6-bit to 4-bit, suitable for text generation tasks.
Large Language Model
Q
mlx-community
561
19
Qwen3 30B A3B Gptq 8bit
Apache-2.0
Qwen3 30B A3B is a large language model that has undergone 8-bit quantization using the GPTQ method, suitable for efficient inference scenarios.
Large Language Model
Transformers

Q
btbtyler09
301
2
Qwen3 235B A22B 4bit
Apache-2.0
This model is a 4-bit quantized version of Qwen/Qwen3-235B-A22B converted to MLX format, suitable for text generation tasks.
Large Language Model
Q
mlx-community
974
6
Qwen3 30B A3B MLX 8bit
Apache-2.0
This model is converted from Qwen/Qwen3-30B-A3B to the MLX format, supporting 8-bit quantization and suitable for text generation tasks.
Large Language Model
Q
lmstudio-community
7,759
6
Qwen3 30B A3B MLX 4bit
Apache-2.0
Qwen3-30B-A3B-MLX-8bit is an 8-bit quantized version converted from Qwen/Qwen3-30B-A3B, optimized for the MLX framework and suitable for text generation tasks.
Large Language Model
Q
lmstudio-community
4,199
19
Qwen3 8B Bf16
Apache-2.0
Qwen3-8B-bf16 is an MLX format model converted from Qwen/Qwen3-8B, supporting text generation tasks.
Large Language Model
Q
mlx-community
1,658
1
Qwen3 30B A3B 8bit
Apache-2.0
Qwen3-30B-A3B-8bit is an MLX format conversion version of the Qwen/Qwen3-30B-A3B model, supporting efficient operation on Apple chips.
Large Language Model
Q
mlx-community
1,553
6
Qwen3 30B A3B 4bit
Apache-2.0
Qwen3-30B-A3B-4bit is a 4-bit quantized version converted from Qwen/Qwen3-30B-A3B, suitable for efficient text generation tasks under the MLX framework.
Large Language Model
Q
mlx-community
2,394
7
Qwen3 32B MLX 4bit
Apache-2.0
This model is a 4-bit quantized version of Qwen3-32B in MLX format, optimized for efficient operation on Apple Silicon devices.
Large Language Model
Q
lmstudio-community
32.14k
3
Huihui Ai.glm 4 32B 0414 Abliterated GGUF
GLM-4-32B-0414-abliterated is a large-scale language model based on the GLM architecture, with a parameter scale of 32B, suitable for text generation tasks.
Large Language Model
H
DevQuasar
623
2
GLM 4 32B 0414 8bit
MIT
This model is an 8-bit quantized MLX format conversion from THUDM/GLM-4-32B-0414, supporting Chinese and English text generation tasks.
Large Language Model Supports Multiple Languages
G
mlx-community
222
4
GLM 4 32B 0414 EXL3
Apache-2.0
GLM-4-32B-0414 is a large-scale language model developed by the THUDM team, based on the GLM architecture, suitable for various text generation tasks.
Large Language Model
G
owentruong
36
2
Qwen2.5 VL 72B Instruct FP8 Dynamic
Apache-2.0
FP8 quantized version of Qwen2.5-VL-72B-Instruct, supporting vision-text input and text output, optimized and released by Neural Magic.
Image-to-Text
Transformers English

Q
parasail-ai
78
1
VL Rethinker 72B 8bit
Apache-2.0
This model is a multimodal vision-language model converted from Qwen2.5-VL-7B-Instruct, supporting 8-bit quantization and suitable for visual question-answering tasks.
Text-to-Image
Transformers English

V
mlx-community
18
0
Gemma 3 27b It Qat 4bit
Other
Gemma 3 27B IT QAT 4bit is an MLX-format model converted from Google's original model, supporting image-to-text tasks.
Image-to-Text
Transformers Other

G
mlx-community
2,200
12
THUDM.GLM 4 32B 0414 GGUF
GLM-4-32B-0414 is a large-scale language model developed by THUDM, with 32 billion parameters, suitable for various text generation tasks.
Large Language Model
T
DevQuasar
13.15k
5
Vora 7B Instruct
VoRA is a vision-language model based on 7B parameters, focusing on image-text-to-text conversion tasks.
Image-to-Text
Transformers

V
Hon-Wong
154
12
Vora 7B Base
VoRA is a vision-language model based on 7B parameters, capable of processing image and text inputs to generate text outputs.
Image-to-Text
Transformers

V
Hon-Wong
62
4
All Hands.openhands Lm 32b V0.1 GGUF
OpenHands LM 32B v0.1 is a 32B-parameter open-source large language model dedicated to the free dissemination of knowledge.
Large Language Model
A
DevQuasar
5,771
2
Deepseek Ai.deepseek V3 0324 GGUF
DeepSeek-V3-0324 is a powerful foundational model focused on text generation tasks, designed to deliver high-quality text generation capabilities.
Large Language Model
D
DevQuasar
2,850
2
Videollama2.1 7B AV CoT
Apache-2.0
VideoLLaMA2.1-7B-AV is a multimodal large language model focused on audio-visual question answering tasks, capable of processing both video and audio inputs to provide high-quality question answering and description generation.
Video-to-Text
Transformers English

V
lym0302
34
0
Qwq 32B NF4
Apache-2.0
This is the 4-bit quantized version of the Qwen/QwQ-32B model, optimized using the BitsAndBytes library, suitable for text generation tasks in resource-constrained environments.
Large Language Model
Transformers English

Q
ginipick
150
27
Olmo2 8B SuperBPE T160k
Apache-2.0
An 8-billion-parameter model featuring the innovative SuperBPE tokenizer, combining subword and super tokens, achieving 30% higher inference efficiency than traditional BPE models.
Large Language Model
Transformers English

O
UW
28
2
Mistral Small 3.1 24b Instruct 2503 Hf
Apache-2.0
Mistral Small 3.1 Instruct 24B is a large language model based on instruction fine-tuning, focusing on text generation tasks.
Large Language Model
Transformers

M
mrfakename
9,416
9
Gemma 3 12b It Codeforces SFT
A large language model fine-tuned on the codeforces-cots dataset based on google/gemma-3-12b-it
Large Language Model
Transformers

G
qgallouedec
43
5
Gemma 3 12b It GGUF
Gemma 3 12B is a large language model that provides a quantized version in GGUF format, suitable for local deployment and use.
Large Language Model
Transformers

G
tensorblock
336
1
Qwq 32B INT8 W8A8
Apache-2.0
INT8 quantized version of QWQ-32B, optimized by reducing the bit-width of weights and activations
Large Language Model
Transformers English

Q
ospatch
590
4
Gemma 3 12b Novision
A text-only version converted from google/gemma-3-12b-it, with visual components removed, focusing on text generation tasks
Large Language Model
Transformers

G
gghfez
86
2
Gemma 3 12b It GGUF
Gemma-3-12b-it is a large language model developed by Google, based on the transformer architecture, focusing on text generation tasks.
Large Language Model
G
second-state
583
1
Gemma 3 27b It Mlx
This is an MLX-converted version of the Google Gemma 3 27B IT model, supporting image-text-to-text tasks.
Image-to-Text
Transformers

G
stephenwalker
24
1
Qwq 32B Bnb 4bit
Apache-2.0
4-bit quantized version of QwQ-32B, optimized using Bitsandbytes technology, suitable for efficient inference in resource-constrained environments
Large Language Model
Transformers

Q
onekq-ai
167
2
- 1
- 2
- 3
Featured Recommended AI Models